5 found
Order:
See also
Samuel Allen Alexander
Ohio State University (PhD)
  1. Pseudo-visibility: A Game Mechanic Involving Willful Ignorance.Samuel Allen Alexander & Arthur Paul Pedersen - 2022 - FLAIRS-35.
    We present a game mechanic called pseudo-visibility for games inhabited by non-player characters (NPCs) driven by reinforcement learning (RL). NPCs are incentivized to pretend they cannot see pseudo-visible players: the training environment simulates an NPC to determine how the NPC would act if the pseudo-visible player were invisible, and penalizes the NPC for acting differently. NPCs are thereby trained to selectively ignore pseudo-visible players, except when they judge that the reaction penalty is an acceptable tradeoff (e.g., a guard might accept (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  2. The Archimedean trap: Why traditional reinforcement learning will probably not yield AGI.Samuel Allen Alexander - 2020 - Journal of Artificial General Intelligence 11 (1):70-85.
    After generalizing the Archimedean property of real numbers in such a way as to make it adaptable to non-numeric structures, we demonstrate that the real numbers cannot be used to accurately measure non-Archimedean structures. We argue that, since an agent with Artificial General Intelligence (AGI) should have no problem engaging in tasks that inherently involve non-Archimedean rewards, and since traditional reinforcement learning rewards are real numbers, therefore traditional reinforcement learning probably will not lead to AGI. We indicate two possible ways (...)
    Direct download  
     
    Export citation  
     
    Bookmark   1 citation  
  3. Reward-Punishment Symmetric Universal Intelligence.Samuel Allen Alexander & Marcus Hutter - 2021 - In AGI.
    Can an agent's intelligence level be negative? We extend the Legg-Hutter agent-environment framework to include punishments and argue for an affirmative answer to that question. We show that if the background encodings and Universal Turing Machine (UTM) admit certain Kolmogorov complexity symmetries, then the resulting Legg-Hutter intelligence measure is symmetric about the origin. In particular, this implies reward-ignoring agents have Legg-Hutter intelligence 0 according to such UTMs.
    Direct download  
     
    Export citation  
     
    Bookmark   1 citation  
  4. Universal Agent Mixtures and the Geometry of Intelligence.Samuel Allen Alexander, David Quarel, Len Du & Marcus Hutter - 2023 - Aistats.
    Inspired by recent progress in multi-agent Reinforcement Learning (RL), in this work we examine the collective intelligent behaviour of theoretical universal agents by introducing a weighted mixture operation. Given a weighted set of agents, their weighted mixture is a new agent whose expected total reward in any environment is the corresponding weighted average of the original agents' expected total rewards in that environment. Thus, if RL agent intelligence is quantified in terms of performance across environments, the weighted mixture's intelligence is (...)
    Direct download  
     
    Export citation  
     
    Bookmark  
  5. Knowledge-of-own-factivity, the definition of surprise, and a solution to the Surprise Examination paradox.Alessandro Aldini, Samuel Allen Alexander & Pierluigi Graziani - 2022 - Cifma.
    Fitch's Paradox and the Paradox of the Knower both make use of the Factivity Principle. The latter also makes use of a second principle, namely the Knowledge-of-Factivity Principle. Both the principle of factivity and the knowledge thereof have been the subject of various discussions, often in conjunction with a third principle known as Closure. In this paper, we examine the well-known Surprise Examination paradox considering both the principles on which this paradox rests and some formal characterisations of the surprise notion, (...)
    Direct download  
     
    Export citation  
     
    Bookmark